On the Design of Loss Functions for Classification: theory, robustness to outliers, and SavageBoost
نویسندگان
چکیده
The machine learning problem of classifier design is studied from the perspective of probability elicitation, in statistics. This shows that the standard approach of proceeding from the specification of a loss, to the minimization of conditional risk is overly restrictive. It is shown that a better alternative is to start from the specification of a functional form for the minimum conditional risk, and derive the loss function. This has various consequences of practical interest, such as showing that 1) the widely adopted practice of relying on convex loss functions is unnecessary, and 2) many new losses can be derived for classification problems. These points are illustrated by the derivation of a new loss which is not convex, but does not compromise the computational tractability of classifier design, and is robust to the contamination of data with outliers. A new boosting algorithm, SavageBoost, is derived for the minimization of this loss. Experimental results show that it is indeed less sensitive to outliers than conventional methods, such as Ada, Real, or LogitBoost, and converges in fewer iterations.
منابع مشابه
SPLBoost: An Improved Robust Boosting Algorithm Based on Self-paced Learning
It is known that Boosting can be interpreted as a gradient descent technique to minimize an underlying loss function. Specifically, the underlying loss being minimized by the traditional AdaBoost is the exponential loss, which is proved to be very sensitive to random noise/outliers. Therefore, several Boosting algorithms, e.g., LogitBoost and SavageBoost, have been proposed to improve the robus...
متن کاملBoosting in the presence of outliers: adaptive classification with non-convex loss functions
This paper examines the role and efficiency of the non-convex loss functions for binary classification problems. In particular, we investigate how to design a simple and effective boosting algorithm that is robust to the outliers in the data. The analysis of the role of a particular non-convex loss for prediction accuracy varies depending on the diminishing tail properties of the gradient of th...
متن کاملRelaxed Clipping: A Global Training Method for Robust Regression and Classification
Robust regression and classification are often thought to require non-convex lossfunctions that prevent scalable, global training. However, such a view neglectsthe possibility of reformulated training methods that can yield practically solvablealternatives. A natural way to make a loss function more robust to outliers isto truncate loss values that exceed a maximum threshold. We...
متن کاملDetermining an Economically Optimal (N,C) Design via Using Loss Functions
In this paper, we introduce a new sampling plan based on the defective proportion of batch. The proposed sampling plan is based on the distribution function of the proportion defective. A continuous loss function is used to quantify deviations between the proportion defective and its acceptance quality level (AQL). For practical purpose, a sensitivity analysis is carried out on the different v...
متن کاملRobustified distance based fuzzy membership function for support vector machine classification
Fuzzification of support vector machine has been utilized to deal with outlier and noise problem. This importance is achieved, by the means of fuzzy membership function, which is generally built based on the distance of the points to the class centroid. The focus of this research is twofold. Firstly, by taking the advantage of robust statistics in the fuzzy SVM, more emphasis on reducing the im...
متن کامل